Some Improved Sample Complexity Bounds in the Probabilistic PAC Learning Model

نویسنده

  • Jun'ichi Takeuchi
چکیده

1 I n t r o d u c t i o n Since Valiant's introduction of the PAC learning model [6] for boolean functions, several extensions of the model to the learning of probability distributions were made. Yamanishi [7] and Kearns and Schapire [3] considered the problem of learning stochastic rules (or probabillstic concepts), which is the problem of learning conditional distributions. Abe and Warmuth [1] looked at the sample as well as computational complexity of learning probability distributions using probabilistic automata as hypotheses. Yamanishi called his learning model the 'stochastic PAC learning model'. Here we use the term the ~probabilistic PAC learning model' to refer to these models collectively. In this paper, we concentrate on the sample complexity aspect of the learning problem for stochastic rules. In particular, we improve upon the best known upper bounds on the sample complexity for learning the important class of stochastic rules called 'the stochastic rules with finite partitioning' (c.f. Yamanishi [7]), with respect to the classic notion of distance among distributions, the KutIback-Leiblsr divergence (KL-divergence, for short). Yamanishi proved an upper bound on the sample complexity of learning stochastic rules with finite partitioning, which was linear in ~ using the property of the MDL estimator. But this result is with respect to a relatively loose notion of distance called the Hellinger distance. 1 With respect to the KullbackLeibler divergence, a similar upper bound had not been obtained. The best known bound to date was of order O(~) due to Abe, Takeuchi, and Warmuth [2]. The quadratic dependence on ~ was due to the fact that Hoeffding's inequality was used to bound the estimation error for lit is well-known that the Kullback-Leibler divergence between two distributions (or stochastic rules) bounds from above the Hellinger distance between the two but the converse does not necessarily hold, even allowing polynomial blow-ups.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Probabilistic Analysis of Learning in Arti cial Neural Networks: The PAC Model and its Variants

There are a number of mathematical approaches to the study of learning and generalization in arti cial neural networks. Here we survey the `probably approximately correct' (PAC) model of learning and some of its variants. These models provide a probabilistic framework for the discussion of generalization and learning. This survey concentrates on the sample complexity questions in these models; ...

متن کامل

Probabilistic Analysis of Learning in Arti cial Neural Networks : The PAC Model and its

There are a number of mathematical approaches to the study of learning and generalization in artiicial neural networks. Here we survey thèprobably approximately correct' (PAC) model of learning and some of its variants. These models provide a probabilistic framework for the discussion of generalization and learning. This survey concentrates on the sample complexity questions in these models; th...

متن کامل

The Sample Complexity and Computational Complexity of Boolean Function Learning

This report surveys some key results on the learning of Boolean functions in a probabilistic model that is a generalization of the well-known ‘PAC’ model. A version of this is to appear as a chapter in a book on Boolean functions, but the report itself is relatively self-contained.

متن کامل

Polynomial Learnability of Stochastic Rules with Respect to the KL-Divergence and Quadratic Distance

We consider the problem of efficient learning of probabilistic concepts (p-concepts) and more generally stochastic rules in the sense defined by Kearns and Schapire [6] and by Yamanishi [18]. Their models extend the PAC-learning model of Valiant [16] to the learning scenario in which the target concept or function is stochastic rather than deterministic as in Valiant’s original model. In this p...

متن کامل

Collaborative PAC Learning

We introduce the collaborative PAC learning model, in which k players attempt to learn the same underlying concept. We ask how much more information is required to learn an accurate classifier for all players simultaneously. We refer to the ratio between the sample complexity of collaborative PAC learning and its non-collaborative (single-player) counterpart as the overhead. We design learning ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 1992